Home
AI Tools
AI Models
MCP
AI NEWS
EN
Model Selection
Tags
Compute-Optimal Training
# Compute-Optimal Training
Cerebras GPT 111M
Apache-2.0
A 111M parameter model in the Cerebras-GPT series, adopting GPT-3 style architecture, trained on The Pile dataset, achieving compute-optimal performance following Chinchilla scaling laws.
Large Language Model
Transformers
English
C
cerebras
5,975
76
Featured Recommended AI Models
Empowering the Future, Your AI Solution Knowledge Base
English
简体中文
繁體中文
にほんご
© 2025
AIbase